perm filename CHAP4[4,KMC]12 blob
sn#030220 filedate 1973-03-21 generic text, type T, neo UTF8
00100 CHAPTER FOUR
00200 PROBLEMS FOR COMPUTER UNDERSTANDING OF NATURAL LANGUAGE
00300 COMMUNICATION IN TELETYPED PSYCHIATRIC INTERVIEWS
00400
00500 Since the behavior being simulated by our
00600 paranoid model is the language-behavior of a paranoid patient in a
00700 psychiatric interview, the model must have an ability to interpret
00800 and respond to natural language input sufficient to demonstrate
00900 language-behavior characteristic of the paranoid mode.
01000 By `natural language` I shall mean ordinary American English such as
01100 is used by readers of this book in everyday conversations. It is
01200 still difficult to be explicit about the processes which enable
01300 hummans to interpret and respond to natural language. Philosophers,
01400 linguists and psychologists have investigated natural language with
01500 various purposes and few useful results. Now attempts are being made
01600 in artificial intelligence to write algorithims which `understand'
01700 natural language expressions.
01800 During the 1960's when
01900 machine processing of natural language was dominated by syntactic
02000 considerations, it became clear that syntactical information alone
02100 was insufficient to comprehend the expressions of ordinary
02200 conversations. The current view is that to understand what is said in
02300 linguistic expressions, syntax and semantics must be combined with
02400 beliefs from an underlying conceptual structure having an ability to
02500 draw inferences. How to achieve this combination efficiently with a
02600 large data-base represents a monumental task for both theory and
02700 implementation. How language
02800 is understood depends on the intentions of the producers and
02900 interpreters in the dialogue. Thus language is understood in
03000 accordance with the participant's view of the situation. Our purpose was to develop a
03100 method for understanding everyday English sufficient for the model to
03200 communicate linguistically in a paranoid way in the circumscribed
03300 situation of a psychiatric interview.
03400 We did not try to construct a general-purpose algorithm which could
03500 understand anything said in English by anybody to anybody in any
03600 dialogue situation. (Does anyone believe it possible?)
03700 We took as a pragmatic measure of "understanding" the ability
03800 of the algorithm to `get the message' of an expression by trying to classify
03900 the imperative or directive intent of the interviewer,i.e.what effect he is
04000 trying to bring about in the interviewee relative to the topic. This
04100 straightforward approach to a complex problem has its drawbacks, as
04200 will be shown, but we strove for a highly individualized idiolect sufficient
04300 to demonstrate paranoid processes of an individual in a particular
04400 situation rather than for a general supra-individual or ideal comprehension
04500 of English. If the language-understanding algorithm interfered with
04600 demonstrating the paranoid processes, we would consider it defective
04700 and insufficient for our purposes. (Insert from PAPER,4,kmc
04800 here)
04900 Some special problems a dialogue algorithm must cope with in a
05000 psychiatric interview will now be discussed.
05100
05200 QUESTIONS
05300
05400 The principal sentence-type used by an interviewer consists
05500 of a question. The usual wh- and yes-no questions must be recognized
05600 by the language-algorithm. In teletyped interviews a question may
05700 sometimes be put in declarative form followed by a question mark as in:
05800 (1) PT.- I LIKE TO GAMBLE ON THE HORSES.
05900 DR.- YOU GAMBLE?
06000
06100 Particularly difficult are `when' questions which require a memory
06200 which can assign each event a beginning, end and a duration. Also
06300 troublesome are questions such as `how often', `how many', i.e. a
06400 `how' followed by a quantifier.
06500 In constructing a simulation of a thought process it is
06600 arbitrary how much information to represent in memory. Should the
06700 model know what is the capital of Alabama? It is trivial to store a lot of facts. We took the position that
06800 the model should know only what we believed it reasonable to know
06900 relevant to a few hundred topics expectable in a psychiatric interview.
07000 Thus the model performs badly when subjected to baiting `exam'
07100 questions designed to test its informational limitations rather than to seek useful
07200 psychiatric information.
07300 IMPERATIVES
07400
07500 Typical imperatives in a psychiatric interview consist of
07600 expressions like:
07700 (2) DR.- TELL ME ABOUT YOURSELF.
07800 (3) DR.- LETS DISCUSS YOUR FAMILY.
07900 Such imperatives are
08000 actually interrogatives to the interviewee about the topics they refer to. Since
08100 the only physical action the model can perform is to `talk' ,
08200 imperatives should be treated as requests for information.
08300 DECLARATIVES
08400
08500 In this category we lump everything else. It includes
08600 greetings, farewells, yes-no type answers, existence assertions and
08700 predications made upon a subject.
08800
08900 AMBIGUITIES
09000
09100 Words have more than one sense, a convenience for human
09200 memories but a struggle for language-analysing algorithms. Consider the
09300 word `bug' in the following expressions:
09400 (4) AM I BUGGING YOU?
09500 (5) AFTER A PERIOD OF HEAVY DRINKING HAVE YOU FELT BUGS ON
09600 YOUR SKIN?
09700 (6) DO YOU THINK THEY PUT A BUG IN YOUR ROOM?
09800 In expression (4) the term
09900 `bug' means to annoy, in (5) it refers to an insect and in (6) it
10000 refers to a microphone used for hidden survellience. Some common words like
10100 `run' have fifty or more common senses. Context must be used to carry
10200 out disambiguation, as described in 00.0. Also we have the advantage
10300 of an idiolect where we can arbitrarily restrict the word senses. One
10400 characteristic of the paranoid mode is that no matter in what sense
10500 the interviewer uses a word, the patient may idiosyncratically
10600 interpret it in some sense relevant to his pathological malevolence
10700 beliefs.
10800
10900 ANAPHORIC REFERENCES
11000 The common anaphoric references consist of the pronouns `it',
11100 `he', `him', `she', `her', `they', `them' as in:
11200 (7) PT.-HORSERACING IS MY HOBBY.
11300 (8) DR.-WHAT DO YOU ENJOY ABOUT IT?
11400 The algorithm must
11500 recognize that the 'it' refers to `horseracing'. More difficult is a
11600 reference more than one I/O pair back in the dialogue as in:
11700 (9) PT.-THE MAFIA IS OUT TO GET ME.
11800 (10) DR.- ARE YOU AFRAID OF THEM?
11900 (11) PT.- MAYBE.
12000 (12) DR.- WHY IS THAT?
12100 The "that" of expression (12) does not refer to
12200 (11) but to the topic of being afraid which the interviewer
12300 introduced in (10). Another pronominal confusion occurs when the
12400 interviewer uses `we' in two senses as in:
12500 (13) DR.- WE WANT YOU TO STAY IN THE HOSPITAL.
12600 (14) PT.- I WANT TO BE DISCHARGED NOW.
12700 (15) DR.- WE ARE NOT COMMUNICATING.
12800 In expression (13) the interviewer
12900 is using "we" to refer to psychiatrists or the hospital staff while
13000 in (15) the term refers to the interviewer and patient.
13100
13200 TOPIC SHIFTS
13300
13400 In the main a psychiatric interviewer is in control of the
13500 interview. When he has gained sufficient information about a topic,
13600 he shifts to a new topic. Naturally the algorithm must detect this
13700 change of topic as in the following:
13800 (16) DR.- HOW DO YOU LIKE THE HOSPITAL?
13900 (17) PT.- ITS NOT HELPING ME TO BE HERE.
14000 (18) DR.- WHAT BROUGHT YOU TO THE HOSPITAL?
14100 (19) PT.- I AM VERY UPSET AND NERVOUS.
14200 (20) DR.- WHAT TENDS TO MAKE YOU NERVOUS?
14300 (22) PT.- JUST BEING AROUND PEOPLE.
14400 (23) DR.- ANYONE IN PARTICULAR?
14500 In (16) and (18) the topic is the hospital. In (20) the
14600 topic changes to causes of the patient's nervous state.
14700 When a topic is introduced by the patient as in (19),
14800 a number of things can be expected to be asked about it. Thus
14900 the algorithm can have ready an expectancy-anaphora list which
15000 allows it to determine whether the topic
15100 introduced by the model is being responded to or whether the interviewer
15200 is continuing with the previous topic.
15300 Topics touched upon previously can be re-introduced
15400 at any point in the interview. The memory of the model is responsible
15500 for knowing what has been discussed.
15600
15700 META-REFERENCES
15800
15900 These are references, not about a topic directly, but about
16000 what has been said about the topic as in:
16100 (24) DR.- WHY ARE YOU IN THE HOSPITAL?
16200 (25) PT.- I SHOULDNT BE HERE.
16300 (26)DR.- WHY DO YOU SAY THAT?
16400 The expression (26 ) is about and meta to expression (25 ).
16500 Sometimes when the patient makes a statement, the doctor replies,
16600 not with a question, but with another statement which constitutes a
16700 rejoinder as in:
16800 (27 ) PT.- I HAVE LOST A LOT OF MONEY GAMBLING.
16900 (28 ) DR.- I GAMBLE QUITE A BIT ALSO.
17000 Here the algorithm should interpret (28 ) as a directive to continue
17100 discussing gambling, not as an indication to question the doctor about
17200 gambling. The one exception to this principle occurs when the algorithm
17300 recognizes a chance to add to its model or representation of the interviewer.
17400 ELLIPSES
17500
17600
17700 In dialogues one finds many ellipses, expressions
17800 from which one or more words are omitted as in:
17900 (29 ) PT.- I SHOULDNT BE HERE.
18000 (30) DR.- WHY NOT?
18100 Here the complete construction must be understood as:
18200 (31) DR.- WHY SHOULD YOU NOT BE HERE?
18300 By saving the previous surface expression and the belief it mapped
18400 into in memory, the algorithm can recognize either what the missing words
18500 are or the concepts they refer to.
18600 The opposite of ellipsis is redundancy which usually provides no
18700 problem since the same thing is being said more than once as in:
18800 (32 ) DR.- LET ME ASK YOU A QUESTION.
18900 If an analysis were required of this expression (it is not
19000 required here since the expression is a sterotype), it would be recognized
19100 that the verb "ask" takes the noun "question" as direct object and
19200 also a question is something that is asked.
19300
19400 SIGNALS
19500
19600 Some fragmentary expressions serve only as directive signals
19700 to proceed as in:
19800 (33) PT.- I WENT TO THE TRACK LAST WEEK.
19900 (34) DR.- AND?
20000 The fragment of (34) requests a continuation of the story
20100 introduced in (33). The common expressions found in interviews are
20200 "and", "so", "go on", "go ahead", "really", etc. If an input expression
20300 cannot be recognized at all, the lowest level fedault condition is
20400 to assume it is a signal and either proceed with the next line in a story under discussion
20500 or if the latter is not the case, begin a new story with a prompting
20600 question or statement.
20700 This strategy can fail as in:
20800 (FIND GOOD EXAMPLE)
20900
21000 IDIOMS
21100
21200 Since so much of conversational language is stereotyped, the task
21300 of recognition is much easier than that of analysis.
21400 This is particularly true of idioms. Either one knows what an idiom means
21500 or one does not. It is usually hopeless to try to decipher what an
21600 idiom means from an analysis of its constituent parts. If the reader
21700 doubts this, let him ponder the following expressions taken from
21800 actual teletyped interviews.
21900 (35) DR.- WHATS EATING YOU?
22000 (36) DR.- YOU SOUND KIND OF PISSED OFF.
22100 (37) DR.- WHAT ARE YOU DRIVING AT?
22200 (38) DR.- ARE YOU PUTTING ME ON?
22300 (39) DR.- WHY ARE THEY AFTER YOU?
22400 (40) DR.- HOW DO YOU GET ALONG WITH THE OTHER PATIENTS?
22500 (41) DR.- HOW DO YOU LIKE YOUR WORK?
22600 (42) DR.- HAVE THEY TRIED TO GET EVEN WITH YOU?
22700 (43) DR.- I CANT KEEP UP WITH YOU.
22800 Understanding idioms is a matter of rote memory. Hence
22900 an algorithm with a large idiom table is required. As each new idiom
23000 appears in teletyped interviews, it should be added to the idiom table
23100 because what happens once can happen again.
23200 One advantage in constructing an idiolect for a model is that
23300 it understands its own idiomatic expressions which tend to be used
23400 by the interviewer if he understands them as in:
23500 (44) PT.- THEY ARE OUT TO GET ME.
23600 (45) DR.- WHAT MAKES YOU THINK THEY ARE OUT TO GET YOU.
23700 The expression (45 ) is really a double idiom in which "out"
23800 means `intend' and "get" means `harm' in this context. Needless to say.
23900 an algorithm which tried to pair off the various meanings of "out" with
24000 the various meanings of "get" would have a hard time of it. But an
24100 algorithm which understands what it itself is capable of saying, should
24200 be able to recognize echoed idioms.
24300
24400 FUZZ TERMS
24500
24600 In this category we group a large number of expressions which
24700 have little or no meaning and therefore can be ignored by the algorithm.
24800 The lower-case expressions in the following are examples of fuzz:
24900 (46) DR.- well now perhaps YOU CAN TELL ME something ABOUT YOUR FAMILY.
25000 (47) DR.- on the other hand I AM INTERESTED IN YOU.
25100 (48) DR.- hey I ASKED YOU A QUESTION.
25200 It is not the case that in order to ignore something one must
25300 recognize explicitly what is ignorable. Since pattern-matching allows
25400 for an `anything' slot in many of its patterns, fuzz is thus easily ignored.
25500
25600 SUBORDINATE CLAUSES
25700
25800 A subordinate clause is a complete statement inside another statement.
25900 It is most frequently introduced by a relative pronoun, indicated in the
26000 following expressions by lower case:
26100 (49) DR.- WAS IT THE UNDERWORLD that PUT YOU HERE?
26200 (50) DR.- WHO ARE THE PEOPLE who UPSET YOU?
26300 (51) DR.- HAS ANYTHING HAPPENED which YOU DONT UNDERSTAND?
26400 The words "whether" and "because" serving as conjunctions are less
26500 frequent. A language-algorithm also must recognize that subordinate clauses
26600 can function as nouns, adjectives, adverbs, and objects of prepositions.
26700
26800 VOCABULARY
26900
27000 How many words should there be in the algorithm's vocabulary?
27100 It is a rare human speaker of English who can recognize 40% of the
27200 415,000 words in the Oxford English Dictionary. In his everyday
27300 conversation an educated person uses perhaps 10,000 words and has
27400 a recognition vocabulary of about 50,000 words. A study of phone
27500 conversations showed that 96 % of the talk employed only 737 words. Of
27600 course the remaining 4% , if not recognized, may be ruinous to the
27700 continuity of a conversation.
27800 In counting the words in 53 teletyped psychiatric interviews,
27900 we found psychiatrists used only 721 words. Since we are familiar with
28000 psychiatric vocabularies and styles of expression, we believed this
28100 language-algorithm could function adequately with a vocabulary
28200 of a few thousand words. There will always be unrecognized words. The
28300 algorithm must be able to continue even if it does not have a particular word
28400 in its vocabulary. This provision represents one great advantage of
28500 pattern-matching over conventional linguistic parsing.
28600 It is not the number of words which creates difficulties but their
28700 combinations. One thousand factorial is still a very large number. Syntactic
28800 and semantic constraints in stereotypes and in analysis reduce this
28900 number to an indefinitely large one.
29000
29100 MISSPELLINGS AND EXTRA CHARACTERS
29200
29300
29400 There is really no good defense against misspellings
29500 in a teletyped interview except having a human monitor retype the correct
29600 versions. Spelling correcting programs are slow, inefficient, and imperfect.
29700 They experience great problems when it is the first character in a word
29800 which is incorrect.
29900 Extra characters sent by the interviewer or by a bad phone
30000 line can be removed by a human monitor.
30100
30200 META VERBS
30300
30400 Certain common verbs such as "think", "feel", "believe", etc
30500 take as their objects a clause as in:
30600 (53) DR.- I THINK YOU ARE RIGHT.
30700 (54) DR.- WHY DO YOU FEEL THE GAMBLING IS CROOKED?
30800 The verb "believe" is peculiar since it can also take as
30900 object a noun or noun phrase as in:
31000 (55) DR.- I BELIEVE YOU.
31100 In expression (54) the conjunction "that" can follow
31200 the word "feel" signifying a subordinate clause. This is not the case
31300 after "believe" in expression (55).
31400
31500 ODD WORDS
31600
31700 These are words which are odd in the context of a
31800 teletyped interview while they are quite natural in the usual vis-a-vis
31900 interview in which the participants communicate through speech. This
32000 should be clear from the following examples in which the odd words
32100 appear in lower case:
32200 (56) DR.-YOU sound CONFUSED.
32300 (57) DR.- DID YOU hear MY LAST QUESTION?
32400 (58) DR.- WOULD YOU come in AND sit down PLEASE?
32500 (59) DR.- CAN YOU say WHO?
32600 (60) DR.- I WILL see YOU AGAIN TOMORROW.
32700
32800
32900 MISUNDERSTANDING
33000
33100 It is not fully recognized bt students of language how often people
33200 misunderstand one another in conversation and yet their
33300 dialogues proceed as if understanding and being understood had taken
33400 place.
33500 The classic story involves three partially deaf men cycling
33600 through the English counrtyside:
33700 FIRST - "WHAT TOWN IS THIS?"
33800 SECOND - "THURSDAY"
33900 THIRD - "ME TOO, LETS STOP AND HAVE A DRINK."
34000 Sometimes a psychiatric interviewer realizes when misunderstanding
34100 occurs and tries to correct it. Other times he simply passes it by. It is
34200 characteristic of the paranoid mode to respond idiosyncratically to
34300 particular word-concepts regardless of what the interviewer is saying:
34400 (FIND GOOD EXAMPLE)
34500
34600 UNUNDERSTANDING
34700 A dialogue algorithm must be prepared for situations
34800 in which it simply does not understand i.e. it cannot arrive at any
34900 interpretation as to what the interviewer is saying. An algorithm should
35000 not be faulted for a lack of facts as in:
35100 (61) DR.- WHO IS THE PRESIDENT OF TURKEY?
35200 wherin the memory does not contain the words "president" and "Turkey".
35300 In this default condition it is simplest to reply:
35400 (62) PT.- I DONT KNOW.
35500 and dangerous to reply:
35600 (63) PT.- COULD YOU REPHRASE THE QUESTION?
35700 because of the horrible loops which can result.
35800 Since the main problem in the default condition of ununderstanding
35900 is how to continue, heuristics can be employed such as asking about the
36000 interviewer's intention as in:
36100 (64) PT.- WHY DO YOU WANT TO KNOW THAT?
36200 or rigidly continuing with a previous topic or introducing a new topic.
36300 These are admittedly desperate measures intended to prompt
36400 the interviewer in directions the algorithm has a better chance of understanding.
36500 Usually it is the interviewer who controls the flow from topic to
36600 topic but there are times, hopefully few, when control must be assumed
36700 by the algorithm.
36800 (Describe language analyzer from Horace PAPER,4,KMC)